107 research outputs found
Approximation Algorithms for Stochastic Boolean Function Evaluation and Stochastic Submodular Set Cover
Stochastic Boolean Function Evaluation is the problem of determining the
value of a given Boolean function f on an unknown input x, when each bit of x_i
of x can only be determined by paying an associated cost c_i. The assumption is
that x is drawn from a given product distribution, and the goal is to minimize
the expected cost. This problem has been studied in Operations Research, where
it is known as "sequential testing" of Boolean functions. It has also been
studied in learning theory in the context of learning with attribute costs. We
consider the general problem of developing approximation algorithms for
Stochastic Boolean Function Evaluation. We give a 3-approximation algorithm for
evaluating Boolean linear threshold formulas. We also present an approximation
algorithm for evaluating CDNF formulas (and decision trees) achieving a factor
of O(log kd), where k is the number of terms in the DNF formula, and d is the
number of clauses in the CNF formula. In addition, we present approximation
algorithms for simultaneous evaluation of linear threshold functions, and for
ranking of linear functions.
Our function evaluation algorithms are based on reductions to the Stochastic
Submodular Set Cover (SSSC) problem. This problem was introduced by Golovin and
Krause. They presented an approximation algorithm for the problem, called
Adaptive Greedy. Our main technical contribution is a new approximation
algorithm for the SSSC problem, which we call Adaptive Dual Greedy. It is an
extension of the Dual Greedy algorithm for Submodular Set Cover due to Fujito,
which is a generalization of Hochbaum's algorithm for the classical Set Cover
Problem. We also give a new bound on the approximation achieved by the Adaptive
Greedy algorithm of Golovin and Krause
Conjunctions of Unate DNF Formulas: Learning and Structure
AbstractA central topic in query learning is to determine which classes of Boolean formulas are efficiently learnable with membership and equivalence queries. We consider the class Rkconsisting of conjunctions ofkunate DNF formulas. This class generalizes the class ofk-clause CNF formulas and the class of unate DNF formulas, both of which are known to be learnable in polynomial time with membership and equivalence queries. We prove that R2can be properly learned with a polynomial number of polynomial-size membership and equivalence queries, but can be properly learned in polynomial time with such queries if and only if P=NP. Thus the barrier to properly learning R2with membership and equivalence queries is computational rather than informational. Few results of this type are known. In our proofs, we use recent results of Hellersteinet al.(1997,J. Assoc. Comput. Mach.43(5), 840–862), characterizing the classes that are polynomial-query learnable, together with work of Bshouty on the monotone dimension of Boolean functions. We extend some of our results to Rkand pose open questions on learning DNF formulas of small monotone dimension. We also prove structural results for Rk. We construct, for any fixedk⩾2, a class of functionsfthat cannot be represented by any formula in Rk, but which cannot be “easily” shown to have this property. More precisely, for any functionfonnvariables in the class, the value offon any polynomial-size set of points in its domain is not a witness thatfcannot be represented by a formula in Rk. Our construction is based on BCH codes
A Local Search Algorithm for the Min-Sum Submodular Cover Problem
We consider the problem of solving the Min-Sum Submodular Cover problem using
local search. The Min-Sum Submodular Cover problem generalizes the NP-complete
Min-Sum Set Cover problem, replacing the input set cover instance with a
monotone submodular set function. A simple greedy algorithm achieves an
approximation factor of 4, which is tight unless P=NP [Streeter and Golovin,
NeurIPS, 2008]. We complement the greedy algorithm with analysis of a local
search algorithm. Building on work of Munagala et al. [ICDT, 2005], we show
that, using simple initialization, a straightforward local search algorithm
achieves a -approximate solution in time
, provided that the monotone submodular set function is
also second-order supermodular. Second-order supermodularity has been shown to
hold for a number of submodular functions of practical interest, including
functions associated with set cover, matching, and facility location. We
present experiments on two special cases of Min-Sum Submodular Cover and find
that the local search algorithm can outperform the greedy algorithm on small
data sets
The Stochastic Score Classification Problem
Consider the following Stochastic Score Classification Problem. A doctor is assessing a patient\u27s risk of developing a certain disease, and can perform n tests on the patient. Each test has a binary outcome, positive or negative. A positive result is an indication of risk, and a patient\u27s score is the total number of positive test results. Test results are accurate. The doctor needs to classify the patient into one of B risk classes, depending on the score (e.g., LOW, MEDIUM, and HIGH risk). Each of these classes corresponds to a contiguous range of scores. Test i has probability p_i of being positive, and it costs c_i to perform. To reduce costs, instead of performing all tests, the doctor will perform them sequentially and stop testing when it is possible to determine the patient\u27s risk category. The problem is to determine the order in which the doctor should perform the tests, so as to minimize expected testing cost. We provide approximation algorithms for adaptive and non-adaptive versions of this problem, and pose a number of open questions
- …